# Local Sparse Global Attention
Lsg Bart Base 16384 Mediasum
A BART model based on LSG technology, optimized for long-sequence summarization tasks, supporting input sequences up to 16,384 tokens in length
Text Generation
Transformers English

L
ccdv
22
2
Lsg Bart Base 4096 Mediasum
BART-base model based on LSG technology, fine-tuned for long text summarization tasks on the MediaSum dataset, supporting sequence processing up to 4096 tokens
Text Generation
Transformers English

L
ccdv
44
0
Lsg Bart Base 4096 Multinews
A BART-base model based on LSG technology, designed for long-text summarization tasks, supporting input sequences of up to 4096 tokens
Text Generation
Transformers English

L
ccdv
26
4
Lsg Bart Base 16384 Pubmed
A long-sequence text summarization model based on the BART architecture, specifically fine-tuned for the PubMed scientific paper dataset, capable of processing input sequences up to 16,384 tokens in length
Text Generation
Transformers English

L
ccdv
22
6
Lsg Bart Base 4096 Pubmed
A long-sequence processing model based on LSG attention mechanism, fine-tuned specifically for scientific paper summarization tasks
Text Generation
Transformers English

L
ccdv
21
3
Lsg Bart Base 16384 Arxiv
A long-sequence processing model based on the BART architecture, optimized for scientific paper summarization tasks, supporting long-text input up to 16,384 tokens
Text Generation
Transformers English

L
ccdv
29
5
Lsg Bart Large 4096
The LSG model is an improved long-sequence processing model based on BART-large, utilizing local + sparse + global attention mechanisms for efficient handling of long-text tasks
Text Generation
Transformers English

L
ccdv
15
0
Featured Recommended AI Models